Hypothesis Testing Procedure
Hypothesis Testing Procedure
Overview
Hypothesis testing is a systematic process for making statistical inferences about population parameters based on sample data. This guide outlines the step-by-step procedure for conducting hypothesis tests.
Step-by-Step Procedure
1. State the Hypotheses
Null Hypothesis ( )
- Represents the status quo or no effect
- Always includes an equality sign (
, , or ) - Example:
(population mean equals 100)
Alternative Hypothesis ( )
- Represents what we aim to find evidence for
- Never includes an equality sign
- Can be one-tailed (
or ) or two-tailed ( ) - Example:
(population mean is greater than 100)
2. Choose Significance Level ( )
- Common values: 0.05 (5%) or 0.01 (1%)
- Represents the probability of Type I error
- Determines the critical value(s)
- Example:
means we're willing to accept a 5% chance of rejecting when it's true
3. Select Appropriate Test
Consider:
- Type of data (quantitative/categorical)
- Number of samples
- Sample size
- Distribution assumptions
- Parameter of interest (mean, proportion, variance, etc.)
Common tests:
- One-sample
-test - Two-sample
-test - ANOVA
- Chi-square test
-test for proportions
4. Calculate Test Statistic
- Formula depends on the chosen test
- Standardizes the observed difference
- Example for one-sample
-test:
5. Determine Critical Value(s)
- Based on:
- Significance level (
) - Degrees of freedom
- Test direction (one/two-tailed)
- Significance level (
- Example: For
and two-tailed -test with :
Critical values =
6. Make Decision
Using Critical Value Method:
- Compare test statistic to critical value(s)
- Reject
if test statistic falls in rejection region - Example: If
, reject
Using P-value Method:
- Calculate p-value (probability of more extreme results)
- Reject
if p-value < - Example: If p-value = 0.03 < 0.05, reject
7. State Conclusion
- Clear statement about the decision
- Include context and practical significance
- Example: "We reject the null hypothesis at
, suggesting that the new teaching method significantly improves test scores."
Common Pitfalls to Avoid
-
Multiple Testing
- Performing multiple tests increases Type I error
- Use appropriate corrections (Bonferroni, etc.)
-
Data Dredging
- Testing multiple hypotheses without pre-specification
- Increases false positive rate
-
Misinterpreting P-values
- P-value is not the probability that
is true - Small p-value doesn't guarantee practical significance
- P-value is not the probability that
-
Ignoring Assumptions
- Each test has specific assumptions
- Violating assumptions can invalidate results
Example: One-Sample t-test
Scenario
Testing if a new teaching method improves test scores (known population mean = 75)
Steps
-
Hypotheses
-
Significance Level
-
Test Selection
- One-sample
-test (comparing mean to known value)
- One-sample
-
Test Statistic
- Sample mean = 78
- Sample SD = 5
- Sample size = 25
-
Critical Value
(one-tailed)
-
Decision
- Reject
-
Conclusion
- "The new teaching method significantly improves test scores (p < 0.05)"